Time Series Classification by Class-Based Mahalanobis Distances

نویسندگان

  • Zoltán Prekopcsák
  • Daniel Lemire
چکیده

To classify time series by nearest neighbor, we need to specify or learn a distance. We consider several variations of the Mahalanobis distance and the related Large Margin Nearest Neighbor Classification (LMNN). We find that the conventional Mahalanobis distance is counterproductive. However, both LMNN and the class-based diagonal Mahalanobis distance are competitive.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Time series classification by class-specific Mahalanobis distance measures

To classify time series by nearest neighbors, we need to specify or learn one or several distance measures. We consider variations of the Mahalanobis distance measures which rely on the inverse covariance matrix of the data. Unfortunately — for time series data — the covariance matrix has often low rank. To alleviate this problem we can either use a pseudoinverse, covariance shrinking or limit ...

متن کامل

metricDTW: local distance metric learning in Dynamic Time Warping

We propose to learn multiple local Mahalanobis distance metrics to perform knearest neighbor (kNN) classification of temporal sequences. Temporal sequences are first aligned by dynamic time warping (DTW); given the alignment path, similarity between two sequences is measured by the DTW distance, which is computed as the accumulated distance between matched temporal point pairs along the alignme...

متن کامل

Values for the Fuzzy -means Classifier in Change Detection for Remote Sensing

We discuss an approach to change detection in digital remotely sensed imagery that relies on the Fuzzy Post Classification Comparison technique. We use the fuzzy -means classifier together with the Mahalanobis distance as the basis for a metric of class membership for a individual pixel. We note that the value of the fuzzy exponent in a fuzzy classifier is based on the ratios of the reciprocals...

متن کامل

Improving Chernoff criterion for classification by using the filled function

Linear discriminant analysis is a well-known matrix-based dimensionality reduction method. It is a supervised feature extraction method used in two-class classification problems. However, it is incapable of dealing with data in which classes have unequal covariance matrices. Taking this issue, the Chernoff distance is an appropriate criterion to measure distances between distributions. In the p...

متن کامل

Learning Expected Hitting Time Distance

Most distance metric learning (DML) approaches focus on learning a Mahalanobis metric for measuring distances between examples. However, for particular feature representations, e.g., histogram features like BOW and SPM, Mahalanobis metric could not model the correlations between these features well. In this work, we define a nonMahalanobis distance for histogram features, via Expected Hitting T...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1010.1526  شماره 

صفحات  -

تاریخ انتشار 2010